Variable selection for Naïve Bayes classification

نویسندگان

چکیده

The Naïve Bayes has proven to be a tractable and efficient method for classification in multivariate analysis. However, features are usually correlated, fact that violates the Bayes’ assumption of conditional independence, may deteriorate method’s performance. Moreover, datasets often characterized by large number features, which complicate interpretation results as well slow down execution. In this paper we propose sparse version classifier is three properties. First, sparsity achieved taking into account correlation structure covariates. Second, different performance measures can used guide selection features. Third, constraints on groups higher interest included. Our proposal leads smart search, yields competitive running times, whereas flexibility terms measure integrated. findings show that, when compared against well-referenced feature approaches, proposed obtains regarding accuracy, times balanced datasets. case with unbalanced (or importance) classes, better compromise between rates classes achieved.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Macro for Naïve Bayes Classification

The supervised classification also known as pattern recognition, discrimination, or supervised learning consists of assigning new cases to one of a set of pre-defined classes given a sample of cases for which the true classes are known. The Naïve Bayes (NB) technique of supervised classification has become increasingly popular in the recent years. Despite its unrealistic assumption that feature...

متن کامل

Semantic Naïve Bayes Classifier for Document Classification

In this paper, we propose a semantic naïve Bayes classifier (SNBC) to improve the conventional naïve Bayes classifier (NBC) by incorporating “document-level” semantic information for document classification (DC). To capture the semantic information from each document, we develop semantic feature extraction and modeling algorithms. For semantic feature extraction, we first apply a log-Bilinear d...

متن کامل

Interpretable Boosted Naïve Bayes Classification

Voting methods such as boosting and bagging provide substantial improvements in classification performance in many problem domains. However, the resulting predictions can prove inscrutable to end-users. This is especially problematic in domains such as medicine, where end-user acceptance often depends on the ability of a classifier to explain its reasoning. Here we propose a variant of the boos...

متن کامل

Image Classification Using Naïve Bayes Classifier

An image classification scheme using Naïve Bayes Classifier is proposed in this paper. The proposed Naive Bayes Classifier-based image classifier can be considered as the maximum a posteriori decision rule. The Naïve Bayes Classifier can produce very accurate classification results with a minimum training time when compared to conventional supervised or unsupervised learning algorithms. Compreh...

متن کامل

Empirical Bayes vs. Fully Bayes Variable Selection

For the problem of variable selection for the normal linear model, fixed penalty selection criteria such as AIC, Cp, BIC and RIC correspond to the posterior modes of a hierarchical Bayes model for various fixed hyperparameter settings. Adaptive selection criteria obtained by empirical Bayes estimation of the hyperparameters have been shown by George and Foster [2000. Calibration and Empirical B...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Computers & Operations Research

سال: 2021

ISSN: ['0305-0548', '1873-765X']

DOI: https://doi.org/10.1016/j.cor.2021.105456